Online Deep Learning Hyperparameter Tuning based on Provenance Analysis

نویسندگان

چکیده

Training Deep Learning (DL) models require adjusting a series of hyperparameters. Although there are several tools to automatically choose the best hyperparameter configuration, user is still main actor take final decision. To decide whether training should continue or try different configurations, needs analyze online hyperparameters most adequate dataset, observing metrics such as accuracy and loss values. Provenance naturally represents data derivation relationships (i.e., transformations, parameter values, etc.), which provide important support in this analysis. Most existing provenance solutions define their own proprietary representations DL users choosing makes analysis interoperability difficult. We present Keras-Prov its extension, named Keras-Prov++, provides an analytical dashboard fine-tuning. Different from current mainstream solutions, captures applications using W3C PROV recommendation, allowing for help deciding on changing hyperparameters’ values after performance validation set. experimental evaluation Keras-Prov++ AlexNet real case study, DenseED, that acts surrogate model solving equations. During analysis, identify scenarios suggest reducing number epochs avoid unnecessary executions fine-tuning learning rate improve accuracy.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Collaborative hyperparameter tuning

Hyperparameter learning has traditionally been a manual task because of the limited number of trials. Today’s computing infrastructures allow bigger evaluation budgets, thus opening the way for algorithmic approaches. Recently, surrogate-based optimization was successfully applied to hyperparameter learning for deep belief networks and to WEKA classifiers. The methods combined brute force compu...

متن کامل

Massively Parallel Hyperparameter Tuning

Modern machine learning models are characterized by large hyperparameter search spaces and prohibitively expensive training costs. For such models, we cannot afford to train candidate models sequentially and wait months before finding a suitable hyperparameter configuration. Hence, we introduce the large-scale regime for parallel hyperparameter tuning, where we need to evaluate orders of magnit...

متن کامل

Efficient Transfer Learning Method for Automatic Hyperparameter Tuning

We propose a fast and effective algorithm for automatic hyperparameter tuning that can generalize across datasets. Our method is an instance of sequential model-based optimization (SMBO) that transfers information by constructing a common response surface for all datasets, similar to Bardenet et al. (2013). The time complexity of reconstructing the response surface at every SMBO iteration in ou...

متن کامل

Online Deep Learning: Learning Deep Neural Networks on the Fly

Deep Neural Networks (DNNs) are typically trained by backpropagation in a batch learning setting, which requires the entire training data to be made available prior to the learning task. This is not scalable for many real-world scenarios where new data arrives sequentially in a stream form. We aim to address an open challenge of “Online Deep Learning” (ODL) for learning DNNs on the fly in an on...

متن کامل

A Grouping Hotel Recommender System Based on Deep Learning and Sentiment Analysis

Recommender systems are important tools for users to identify their preferred items and for businesses to improve their products and services. In recent years, the use of online services for selection and reservation of hotels have witnessed a booming growth. Customer’ reviews have replaced the word of mouth marketing, but searching hotels based on user priorities is more time-consuming. This s...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Journal of Information and Data Management

سال: 2021

ISSN: ['2178-7107']

DOI: https://doi.org/10.5753/jidm.2021.1924